Using Kullback-Leibler Divergence to Model Opponents in Poker

نویسندگان

  • Jiajia Zhang
  • Xuan Wang
  • Lin Yao
  • Jingpeng Li
  • Xuedong Shen
چکیده

Opponent modeling is an essential approach for building competitive computer agents in imperfect information games. This paper presents a novel approach to develop opponent modeling techniques. The approach applies neural networks which are separately trained on different dataset to build Kmodel clustering opponent models. KullbackLeibler (KL) divergence is used to exploit a safety mode on opponent modeling. Given a parameter d that controls the max divergence between a model’s centre point and the units belong to it, the approach is proved to provide a lower bound of expected payoff which is above the minimax payoff for correctly clustered players. Even for the players that are incorrectly clustered, the lower bound can also be unlimited approximated with sufficient history data. In our experiments, agent with the novel model shows an improved classification efficiency of opponent modeling comparing with relative researches. And also, the new agent performs better when playing against poker agent HITSZ_CS_13 which participate Annual Computer Poker Competition of 2013.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Model Confidence Set Based on Kullback-Leibler Divergence Distance

Consider the problem of estimating true density, h(.) based upon a random sample X1,…, Xn. In general, h(.)is approximated using an appropriate in some sense, see below) model fƟ(x). This article using Vuong's (1989) test along with a collection of k(> 2) non-nested models constructs a set of appropriate models, say model confidence set, for unknown model h(.).Application of such confide...

متن کامل

Comparison of Kullback-Leibler, Hellinger and LINEX with Quadratic Loss Function in Bayesian Dynamic Linear Models: Forecasting of Real Price of Oil

In this paper we intend to examine the application of Kullback-Leibler, Hellinger and LINEX loss function in Dynamic Linear Model using the real price of oil for 106 years of data from 1913 to 2018 concerning the asymmetric problem in filtering and forecasting. We use DLM form of the basic Hoteling Model under Quadratic loss function, Kullback-Leibler, Hellinger and LINEX trying to address the ...

متن کامل

Using Kullback-Leibler distance for performance evaluation of search designs

This paper considers the search problem, introduced by Srivastava cite{Sr}. This is a model discrimination problem. In the context of search linear models, discrimination ability of search designs has been studied by several researchers. Some criteria have been developed to measure this capability, however, they are restricted in a sense of being able to work for searching only one possibl...

متن کامل

Information Measures via Copula Functions

In applications of differential geometry to problems of parametric inference, the notion of divergence is often used to measure the separation between two parametric densities. Among them, in this paper, we will verify measures such as Kullback-Leibler information, J-divergence, Hellinger distance, -Divergence, … and so on. Properties and results related to distance between probability d...

متن کامل

Evaluating the Improvement of Partial Discharge Localization Accuracy Using Frequency Response Assurance Criterion

Partial Discharge (PD) is the most important source of insulation degradation in power transformers. In order to prevent catastrophic failures in transformers, PDs need to be located as soon as possible so that maintenance measures can be taken in time. Due to the structural complexity of windings, locating the PD source inside a transformer winding is not a simple task. In this paper, the effi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014